Goto

Collaborating Authors

 meta-continual learning


Generative vs. Discriminative: Rethinking The Meta-Continual Learning

Neural Information Processing Systems

Deep neural networks have achieved human-level capabilities in various learning tasks. However, they generally lose performance in more realistic scenarios like learning in a continual manner. In contrast, humans can incorporate their prior knowledge to learn new concepts efficiently without forgetting older ones. In this work, we leverage meta-learning to encourage the model to learn how to learn continually. Inspired by human concept learning, we develop a generative classifier that efficiently uses data-driven experience to learn new concepts even from few samples while being immune to forgetting. Along with cognitive and theoretical insights, extensive experiments on standard benchmarks demonstrate the effectiveness of the proposed method. The ability to remember all previous concepts, with negligible computational and structural overheads, suggests that generative models provide a natural way for alleviating catastrophic forgetting, which is a major drawback of discriminative models.


Online Fast Adaptation and Knowledge Accumulation (OSAKA): a New Approach to Continual Learning

Neural Information Processing Systems

Continual learning agents experience a stream of (related) tasks. The main challenge is that the agent must not forget previous tasks and also adapt to novel tasks in the stream. We are interested in the intersection of two recent continual-learning scenarios. In meta-continual learning, the model is pre-trained using meta-learning to minimize catastrophic forgetting of previous tasks. In continual-meta learning, the aim is to train agents for faster remembering of previous tasks through adaptation.


A A unifying framework Data Distribution Model for Fast Weights Slow Weights Updates Evaluation Supervised Learning S, Q C f

Neural Information Processing Systems

For readability, we omit OSAKA pre-training. Replay-based methods store representative samples from the past, either in their original form (e.g., rehearsal Most prior-based methods rely on task boundaries. Since non-stationary data distributions breaks the i.i.d assumption for The update is computed from a parametric combination of the gradient of the current and previous task. Despite that, meta-continual learning is actively researched [61, 6]. Bayesian change-point detection scheme to identify whether a task has changed.




Generative vs. Discriminative: Rethinking The Meta-Continual Learning

Neural Information Processing Systems

Deep neural networks have achieved human-level capabilities in various learning tasks. However, they generally lose performance in more realistic scenarios like learning in a continual manner. In contrast, humans can incorporate their prior knowledge to learn new concepts efficiently without forgetting older ones. In this work, we leverage meta-learning to encourage the model to learn how to learn continually. Inspired by human concept learning, we develop a generative classifier that efficiently uses data-driven experience to learn new concepts even from few samples while being immune to forgetting. Along with cognitive and theoretical insights, extensive experiments on standard benchmarks demonstrate the effectiveness of the proposed method.


Online Fast Adaptation and Knowledge Accumulation (OSAKA): a New Approach to Continual Learning

Neural Information Processing Systems

Continual learning agents experience a stream of (related) tasks. The main challenge is that the agent must not forget previous tasks and also adapt to novel tasks in the stream. We are interested in the intersection of two recent continual-learning scenarios. In meta-continual learning, the model is pre-trained using meta-learning to minimize catastrophic forgetting of previous tasks. In continual-meta learning, the aim is to train agents for faster remembering of previous tasks through adaptation.


Online Fast Adaptation and Knowledge Accumulation: a New Approach to Continual Learning

#artificialintelligence

Learning from non-stationary data remains a great challenge for machine learning. Continual learning addresses this problem in scenarios where the learning agent faces a stream of changing tasks. In these scenarios, the agent is expected to retain its highest performance on previous tasks without revisiting them while adapting well to the new tasks. Two new recent continual-learning scenarios have been proposed. In meta-continual learning, the model is pre-trained to minimize catastrophic forgetting when trained on a sequence of tasks. In continual-meta learning, the goal is faster remembering, i.e., focusing on how quickly the agent recovers performance rather than measuring the agent's performance without any adaptation.